Efficient crowdsourcing of unknown experts using bounded multi-armed bandits

نویسندگان

  • Long Tran-Thanh
  • Sebastian Stein
  • Alex Rogers
  • Nicholas R. Jennings
چکیده

Increasingly, organisations flexibly outsource work on a temporary basis to a global audience of workers. This so-called crowdsourcing has been applied successfully to a range of tasks, from translating text and annotating images, to collecting information during crisis situations and hiring skilled workers to build complex software. While traditionally these tasks have been small and could be completed by non-professionals, organisations are now starting to crowdsource larger, more complex tasks to experts in their respective fields. These tasks include, for example, software development and testing, web design and product marketing. While this emerging expert crowdsourcing offers flexibility and potentially lower costs, it also raises new challenges, as workers can be highly heterogeneous, both in their costs and in the quality of the work they produce. Specifically, the utility of each outsourced task is uncertain and can vary significantly between distinct workers and even between subsequent tasks assigned to the same worker. Furthermore, in realistic settings, workers have limits on the amount of work they can perform and the employer will have a fixed budget for paying workers. Given this uncertainty and the relevant constraints, the objective of the employer is to assign tasks to workers in order to maximise the overall utility achieved. To formalise this expert crowdsourcing problem, we introduce a novel multi-armed bandit (MAB) model, the bounded MAB. Furthermore, we develop an algorithm to solve it efficiently, called bounded ε-first, which proceeds in two stages: exploration and exploitation. During exploration, it first uses εB of its total budget B to learn estimates of the workers’ quality characteristics. Then, during exploitation, it uses the remaining (1 − ε) B to maximise the total utility based on those estimates. Using this technique allows us to derive an O (

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficient Crowdsourcing of Unknown Experts using Multi-Armed Bandits

We address the expert crowdsourcing problem, in which an employer wishes to assign tasks to a set of available workers with heterogeneous working costs. Critically, as workers produce results of varying quality, the utility of each assigned task is unknown and can vary both between workers and individual tasks. Furthermore, in realistic settings, workers are likely to have limits on the number ...

متن کامل

From Bandits to Experts: A Tale of Domination and Independence

We consider the partial observability model for multi-armed bandits, introduced by Mannor and Shamir [11]. Our main result is a characterization of regret in the directed observability model in terms of the dominating and independence numbers of the observability graph. We also show that in the undirected case, the learner can achieve optimal regret without even accessing the observability grap...

متن کامل

From Bandits to Experts: A Tale of Domination and Independence

We consider the partial observability model for multi-armed bandits, introducedby Mannor and Shamir [11]. Our main result is a characterization of regret inthe directed observability model in terms of the dominating and independencenumbers of the observability graph. We also show that in the undirected case, thelearner can achieve optimal regret without even accessing the observ...

متن کامل

Global Bandits

Standard multi-armed bandits model decision problems in which the consequences of each action choice are unknown and independent of each other. But in a wide variety of decision problems – from drug dosage to dynamic pricing – the consequences (rewards) of different actions are correlated, so that selecting one action provides information about the consequences (rewards) of other actions as wel...

متن کامل

Bandits with Knapsacks: Dynamic procurement for crowdsourcing∗

In a basic version of the dynamic procurement problem, the algorithm has a budget B to spend, and is facing n agents (potential sellers) that are arriving sequentially. The algorithm offers a take-it-or-leave-it price to each arriving seller; the sellers value for an item is an independent sample from some fixed (but unknown) distribution. The goal is to maximize the number of items bought. Thi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Artif. Intell.

دوره 214  شماره 

صفحات  -

تاریخ انتشار 2014